Adaptive Hybrid Storage Format for Sparse Matrix–Vector Multiplication on Multi-Core SIMD CPUs

نویسندگان

چکیده

Optimizing sparse matrix–vector multiplication (SpMV) is challenging due to the non-uniform distribution of non-zero elements matrix. The best-performing SpMV format changes depending on input matrix and underlying architecture, there no “one-size-fit-for-all” format. A hybrid scheme combining multiple storage formats allows one choose an appropriate use for target hardware. However, existing approaches are inadequate utilizing SIMD cores modern multi-core CPUs with SIMDs, it remains unclear how best mix different a given This paper presents new matrices, specifically targeting SIMDs. Our approach partitions into two segmentations based regularities memory access pattern, where each segmentation stored in suitable its patterns. Unlike prior schemes that rely user determine data partition among formats, we employ machine learning build predictive model automatically threshold per basis. first trained off line, can be applied any new, unseen We apply our 956 matrices evaluate performance three distinct CPU platforms: 72-core Intel Knights Landing (KNL) CPU, 128-core AMD EPYC 64-core Phytium ARMv8 CPU. Experimental results show scheme, combined model, outperforms alternative by 2.9%, 17.5% 16% average KNL, AMD, Phytium, respectively.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A Unified Sparse Matrix Data Format for Efficient General Sparse Matrix-Vector Multiplication on Modern Processors with Wide SIMD Units

Sparse matrix-vector multiplication (spMVM) is the most time-consuming kernel in many numerical algorithms and has been studied extensively on all modern processor and accelerator architectures. However, the optimal sparse matrix data storage format is highly hardware-specific, which could become an obstacle when using heterogeneous systems. Also, it is as yet unclear how the wide single instru...

متن کامل

Sparse Matrix Storage Format

Operations on Sparse Matrices are the key computational kernels in many scientific and engineering applications. They are characterized with poor substantiated performance. It is not uncommon for microprocessors to gain only 10-20% of their peak floating-point performance when doing sparse matrix computations even when special vector processors have been added as coprocessor facilities. In this...

متن کامل

Block-Row Sparse Matrix-Vector Multiplication on SIMD Machines

The irregular nature of the data structures required to efficiently store arbitrary sparse matrices and the architectural constraints of a SIMD computer make it difficult to design an algorithm that can efficiently multiply an arbitrary sparse matrix by a vector. A new ‘‘block-row’’ algorithm is proposed. It allows the ‘‘regularity’’ of a data structure with a row-major mapping to be varied by ...

متن کامل

A hybrid format for better performance of sparse matrix-vector multiplication on a GPU

In this paper, we present a new sparse matrix data format that leads to improved memory coalescing and more efficient sparse matrix-vector multiplication (SpMV) for a wide range of problems on high throughput architectures such as a graphics processing unit (GPU). The sparse matrix structure is constructed by sorting the rows based on the row length (defined as the number of non-zero elements i...

متن کامل

Chemical Kinetics on Multi-core SIMD Architectures

Chemical kinetics modeling accounts for a significant portion of the computational time of atmospheric models. Effective application of multiple levels of heterogeneous parallelism can significantly reduce computational time, but implementation on emerging multi-core technologies can be prohibitively difficult. We introduce an approach for chemical kinetics modeling on multi-core SIMD architect...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Applied sciences

سال: 2022

ISSN: ['2076-3417']

DOI: https://doi.org/10.3390/app12199812